var runtime.memstats
205 uses
runtime (current package)
heapdump.go#L714: sysFree(unsafe.Pointer(&tmpbuf[0]), uintptr(len(tmpbuf)), &memstats.other_sys)
heapdump.go#L732: sysFree(unsafe.Pointer(&tmpbuf[0]), uintptr(len(tmpbuf)), &memstats.other_sys)
heapdump.go#L735: p := sysAlloc(n, &memstats.other_sys)
iface.go#L67: m = (*itab)(persistentalloc(unsafe.Sizeof(itab{})+uintptr(len(inter.mhdr)-1)*goarch.PtrSize, 0, &memstats.other_sys))
malloc.go#L439: memstats.by_size[i].size = uint32(class_to_size[i])
malloc.go#L642: v = h.arena.alloc(n, heapArenaBytes, &memstats.heap_sys)
malloc.go#L752: r = (*heapArena)(h.heapArenaAlloc.alloc(unsafe.Sizeof(*r), goarch.PtrSize, &memstats.gcMiscSys))
malloc.go#L754: r = (*heapArena)(persistentalloc(unsafe.Sizeof(*r), goarch.PtrSize, &memstats.gcMiscSys))
malloc.go#L766: newArray := (*notInHeap)(persistentalloc(size, goarch.PtrSize, &memstats.gcMiscSys))
malloc.go#L947: return persistentalloc(size, align, &memstats.other_sys)
malloc.go#L1437: persistent.base = (*notInHeap)(sysAlloc(persistentChunkSize, &memstats.other_sys))
malloc.go#L1462: if sysStat != &memstats.other_sys {
malloc.go#L1464: memstats.other_sys.add(-int64(size))
mbitmap.go#L1032: debugPtrmask.data = (*byte)(persistentalloc(1<<20, 1, &memstats.other_sys))
mbitmap.go#L1570: x := (*[1 << 30]byte)(persistentalloc(n+1, 1, &memstats.buckhash_sys))[:n+1]
mcache.go#L177: stats := memstats.heapStats.acquire()
mcache.go#L185: memstats.heapStats.release()
mcache.go#L217: stats := memstats.heapStats.acquire()
mcache.go#L220: memstats.heapStats.release()
mcache.go#L245: stats := memstats.heapStats.acquire()
mcache.go#L247: memstats.heapStats.release()
mcache.go#L268: stats := memstats.heapStats.acquire()
mcache.go#L271: memstats.heapStats.release()
mcheckmark.go#L46: bitmap = (*checkmarksMap)(persistentalloc(unsafe.Sizeof(*bitmap), 0, &memstats.gcMiscSys))
metrics.go#L168: hist.counts[0] = atomic.Load64(&memstats.gcPauseDist.underflow)
metrics.go#L169: for i := range memstats.gcPauseDist.counts {
metrics.go#L170: hist.counts[i+1] = atomic.Load64(&memstats.gcPauseDist.counts[i])
metrics.go#L388: memstats.heapStats.read(&a.heapStatsDelta)
metrics.go#L430: a.stacksSys = memstats.stacks_sys.load()
metrics.go#L431: a.buckHashSys = memstats.buckhash_sys.load()
metrics.go#L432: a.gcMiscSys = memstats.gcMiscSys.load()
metrics.go#L433: a.otherSys = memstats.other_sys.load()
metrics.go#L435: a.gcCyclesDone = uint64(memstats.numgc)
metrics.go#L436: a.gcCyclesForced = uint64(memstats.numforcedgc)
metrics.go#L440: a.mSpanSys = memstats.mspan_sys.load()
metrics.go#L442: a.mCacheSys = memstats.mcache_sys.load()
mfinal.go#L92: finc = (*finblock)(persistentalloc(_FinBlockSize, 0, &memstats.gcMiscSys))
mgc.go#L181: memstats.enablegc = true // now that runtime is initialized, GC is okay
mgc.go#L546: if !memstats.enablegc || panicking != 0 || gcphase != _GCoff {
mgc.go#L560: lastgc := int64(atomic.Load64(&memstats.last_gc_nanotime))
mgc.go#L727: memstats.gcPauseDist.record(now - work.pauseStart)
mgc.go#L875: memstats.gcPauseDist.record(now - work.pauseStart)
mgc.go#L976: memstats.last_heap_inuse = memstats.heap_inuse
mgc.go#L989: memstats.gcPauseDist.record(now - work.pauseStart)
mgc.go#L990: atomic.Store64(&memstats.last_gc_unix, uint64(unixNow)) // must be Unix time to make sense to user
mgc.go#L991: atomic.Store64(&memstats.last_gc_nanotime, uint64(now)) // monotonic time for us
mgc.go#L992: memstats.pause_ns[memstats.numgc%uint32(len(memstats.pause_ns))] = uint64(work.pauseNS)
mgc.go#L993: memstats.pause_end[memstats.numgc%uint32(len(memstats.pause_end))] = uint64(unixNow)
mgc.go#L994: memstats.pause_total_ns += uint64(work.pauseNS)
mgc.go#L1007: memstats.gc_cpu_fraction = float64(work.totaltime) / float64(totalCpu)
mgc.go#L1014: memstats.numforcedgc++
mgc.go#L1019: memstats.numgc++
mgc.go#L1068: util := int(memstats.gc_cpu_fraction * 100)
mgc.go#L1072: print("gc ", memstats.numgc,
mgcscavenge.go#L104: return memstats.heap_sys.load() - atomic.Load64(&memstats.heap_released)
mgcscavenge.go#L133: retainedGoal := uint64(float64(memstats.last_heap_inuse) * goalRatio)
mgcscavenge.go#L478: atomic.Load64(&memstats.heap_released)>>10, " KiB total, ",
mgcscavenge.go#L479: (atomic.Load64(&memstats.heap_inuse)*100)/heapRetained(), "% util",
mgcscavenge.go#L780: atomic.Xadd64(&memstats.heap_released, nbytes)
mgcscavenge.go#L783: stats := memstats.heapStats.acquire()
mgcscavenge.go#L786: memstats.heapStats.release()
mgcsweep.go#L668: stats := memstats.heapStats.acquire()
mgcsweep.go#L670: memstats.heapStats.release()
mgcsweep.go#L715: stats := memstats.heapStats.acquire()
mgcsweep.go#L718: memstats.heapStats.release()
mheap.go#L503: sp.array = sysAlloc(uintptr(n)*goarch.PtrSize, &memstats.other_sys)
mheap.go#L515: sysFree(unsafe.Pointer(&oldAllspans[0]), uintptr(cap(oldAllspans))*unsafe.Sizeof(oldAllspans[0]), &memstats.other_sys)
mheap.go#L701: h.spanalloc.init(unsafe.Sizeof(mspan{}), recordspan, unsafe.Pointer(h), &memstats.mspan_sys)
mheap.go#L702: h.cachealloc.init(unsafe.Sizeof(mcache{}), nil, nil, &memstats.mcache_sys)
mheap.go#L703: h.specialfinalizeralloc.init(unsafe.Sizeof(specialfinalizer{}), nil, nil, &memstats.other_sys)
mheap.go#L704: h.specialprofilealloc.init(unsafe.Sizeof(specialprofile{}), nil, nil, &memstats.other_sys)
mheap.go#L705: h.specialReachableAlloc.init(unsafe.Sizeof(specialReachable{}), nil, nil, &memstats.other_sys)
mheap.go#L706: h.arenaHintAlloc.init(unsafe.Sizeof(arenaHint{}), nil, nil, &memstats.other_sys)
mheap.go#L723: h.pages.init(&h.lock, &memstats.gcMiscSys)
mheap.go#L1280: atomic.Xadd64(&memstats.heap_released, -int64(scav))
mheap.go#L1284: atomic.Xadd64(&memstats.heap_inuse, int64(nbytes))
mheap.go#L1288: memstats.heap_sys.add(-int64(nbytes))
mheap.go#L1291: stats := memstats.heapStats.acquire()
mheap.go#L1304: memstats.heapStats.release()
mheap.go#L1360: print("runtime: out of memory: cannot allocate ", ask, "-byte block (", memstats.heap_sys, " in use)\n")
mheap.go#L1376: sysMap(unsafe.Pointer(h.curArena.base), size, &memstats.heap_sys)
mheap.go#L1378: atomic.Xadd64(&memstats.heap_released, int64(size))
mheap.go#L1379: stats := memstats.heapStats.acquire()
mheap.go#L1381: memstats.heapStats.release()
mheap.go#L1404: sysMap(unsafe.Pointer(v), nBase-v, &memstats.heap_sys)
mheap.go#L1412: atomic.Xadd64(&memstats.heap_released, int64(nBase-v))
mheap.go#L1413: stats := memstats.heapStats.acquire()
mheap.go#L1415: memstats.heapStats.release()
mheap.go#L1490: atomic.Xadd64(&memstats.heap_inuse, -int64(nbytes))
mheap.go#L1494: memstats.heap_sys.add(int64(nbytes))
mheap.go#L1497: stats := memstats.heapStats.acquire()
mheap.go#L1508: memstats.heapStats.release()
mheap.go#L2100: result = (*gcBitsArena)(sysAlloc(gcBitsChunkBytes, &memstats.gcMiscSys))
mprof.go#L174: b := (*bucket)(persistentalloc(size, 0, &memstats.buckhash_sys))
mprof.go#L208: buckhash = (*[buckHashSize]*bucket)(sysAlloc(unsafe.Sizeof(*buckhash), &memstats.buckhash_sys))
mspanset.go#L105: newSpine := persistentalloc(newCap*goarch.PtrSize, cpu.CacheLineSize, &memstats.gcMiscSys)
mspanset.go#L286: return (*spanSetBlock)(persistentalloc(unsafe.Sizeof(spanSetBlock{}), cpu.CacheLineSize, &memstats.gcMiscSys))
mstats.go#L104: var memstats mstats
mstats.go#L380: if offset := unsafe.Offsetof(memstats.heapStats); offset%8 != 0 {
mstats.go#L384: if offset := unsafe.Offsetof(memstats.gcPauseDist); offset%8 != 0 {
mstats.go#L415: stats.Alloc = memstats.alloc
mstats.go#L416: stats.TotalAlloc = memstats.total_alloc
mstats.go#L417: stats.Sys = memstats.sys
mstats.go#L418: stats.Mallocs = memstats.nmalloc
mstats.go#L419: stats.Frees = memstats.nfree
mstats.go#L420: stats.HeapAlloc = memstats.alloc
mstats.go#L421: stats.HeapSys = memstats.heap_sys.load()
mstats.go#L438: stats.HeapIdle = memstats.heap_sys.load() - memstats.heap_inuse
mstats.go#L439: stats.HeapInuse = memstats.heap_inuse
mstats.go#L440: stats.HeapReleased = memstats.heap_released
mstats.go#L441: stats.HeapObjects = memstats.heap_objects
mstats.go#L442: stats.StackInuse = memstats.stacks_inuse
mstats.go#L445: stats.StackSys = memstats.stacks_inuse + memstats.stacks_sys.load()
mstats.go#L446: stats.MSpanInuse = memstats.mspan_inuse
mstats.go#L447: stats.MSpanSys = memstats.mspan_sys.load()
mstats.go#L448: stats.MCacheInuse = memstats.mcache_inuse
mstats.go#L449: stats.MCacheSys = memstats.mcache_sys.load()
mstats.go#L450: stats.BuckHashSys = memstats.buckhash_sys.load()
mstats.go#L454: stats.GCSys = memstats.gcMiscSys.load() + memstats.gcWorkBufInUse + memstats.gcProgPtrScalarBitsInUse
mstats.go#L455: stats.OtherSys = memstats.other_sys.load()
mstats.go#L457: stats.LastGC = memstats.last_gc_unix
mstats.go#L458: stats.PauseTotalNs = memstats.pause_total_ns
mstats.go#L459: stats.PauseNs = memstats.pause_ns
mstats.go#L460: stats.PauseEnd = memstats.pause_end
mstats.go#L461: stats.NumGC = memstats.numgc
mstats.go#L462: stats.NumForcedGC = memstats.numforcedgc
mstats.go#L463: stats.GCCPUFraction = memstats.gc_cpu_fraction
mstats.go#L475: if l := len(memstats.by_size); l < bySizeLen {
mstats.go#L479: stats.BySize[i].Size = memstats.by_size[i].size
mstats.go#L480: stats.BySize[i].Mallocs = memstats.by_size[i].nmalloc
mstats.go#L481: stats.BySize[i].Frees = memstats.by_size[i].nfree
mstats.go#L498: if cap(p) < len(memstats.pause_ns)+3 {
mstats.go#L505: n := memstats.numgc
mstats.go#L506: if n > uint32(len(memstats.pause_ns)) {
mstats.go#L507: n = uint32(len(memstats.pause_ns))
mstats.go#L516: j := (memstats.numgc - 1 - i) % uint32(len(memstats.pause_ns))
mstats.go#L517: p[i] = memstats.pause_ns[j]
mstats.go#L518: p[n+i] = memstats.pause_end[j]
mstats.go#L521: p[n+n] = memstats.last_gc_unix
mstats.go#L522: p[n+n+1] = uint64(memstats.numgc)
mstats.go#L523: p[n+n+2] = memstats.pause_total_ns
mstats.go#L542: memstats.mcache_inuse = uint64(mheap_.cachealloc.inuse)
mstats.go#L543: memstats.mspan_inuse = uint64(mheap_.spanalloc.inuse)
mstats.go#L544: memstats.sys = memstats.heap_sys.load() + memstats.stacks_sys.load() + memstats.mspan_sys.load() +
mstats.go#L545: memstats.mcache_sys.load() + memstats.buckhash_sys.load() + memstats.gcMiscSys.load() +
mstats.go#L546: memstats.other_sys.load()
mstats.go#L555: memstats.alloc = 0
mstats.go#L556: memstats.total_alloc = 0
mstats.go#L557: memstats.nmalloc = 0
mstats.go#L558: memstats.nfree = 0
mstats.go#L559: for i := 0; i < len(memstats.by_size); i++ {
mstats.go#L560: memstats.by_size[i].nmalloc = 0
mstats.go#L561: memstats.by_size[i].nfree = 0
mstats.go#L565: memstats.heapStats.unsafeRead(&consStats)
mstats.go#L569: memstats.nmalloc += uint64(consStats.largeAllocCount)
mstats.go#L571: memstats.nfree += uint64(consStats.largeFreeCount)
mstats.go#L578: memstats.nmalloc += a
mstats.go#L579: memstats.by_size[i].nmalloc = a
mstats.go#L584: memstats.nfree += f
mstats.go#L585: memstats.by_size[i].nfree = f
mstats.go#L589: memstats.nfree += uint64(consStats.tinyAllocCount)
mstats.go#L590: memstats.nmalloc += uint64(consStats.tinyAllocCount)
mstats.go#L593: memstats.total_alloc = totalAlloc
mstats.go#L594: memstats.alloc = totalAlloc - totalFree
mstats.go#L595: memstats.heap_objects = memstats.nmalloc - memstats.nfree
mstats.go#L597: memstats.stacks_inuse = uint64(consStats.inStacks)
mstats.go#L598: memstats.gcWorkBufInUse = uint64(consStats.inWorkBufs)
mstats.go#L599: memstats.gcProgPtrScalarBitsInUse = uint64(consStats.inPtrScalarBits)
mstats.go#L602: memstats.sys += memstats.stacks_inuse + memstats.gcWorkBufInUse + memstats.gcProgPtrScalarBitsInUse
mstats.go#L616: if memstats.heap_inuse != uint64(consStats.inHeap) {
mstats.go#L617: print("runtime: heap_inuse=", memstats.heap_inuse, "\n")
mstats.go#L621: if memstats.heap_released != uint64(consStats.released) {
mstats.go#L622: print("runtime: heap_released=", memstats.heap_released, "\n")
mstats.go#L626: globalRetained := memstats.heap_sys.load() - memstats.heap_released
netpoll.go#L623: mem := persistentalloc(n*pdSize, 0, &memstats.other_sys)
os_linux.go#L189: stack := sysAlloc(stacksize, &memstats.stacks_sys)
proc.go#L675: lockInit(&memstats.heapStats.noPLock, lockRankLeafRank)
stack.go#L359: v := sysAlloc(uintptr(n), &memstats.stacks_sys)
stack.go#L463: sysFree(v, n, &memstats.stacks_sys)
trace.go#L366: sysFree(unsafe.Pointer(buf), unsafe.Sizeof(*buf.ptr()), &memstats.other_sys)
trace.go#L655: buf = traceBufPtr(sysAlloc(unsafe.Sizeof(traceBuf{}), &memstats.other_sys))
trace.go#L962: block := (*traceAllocBlock)(sysAlloc(unsafe.Sizeof(traceAllocBlock{}), &memstats.other_sys))
trace.go#L980: sysFree(unsafe.Pointer(block), unsafe.Sizeof(traceAllocBlock{}), &memstats.other_sys)